Goto

Collaborating Authors

 control rule


Using machine learning to inform harvest control rule design in complex fishery settings

Montealegre-Mora, Felipe, Boettiger, Carl, Walters, Carl J., Cahill, Christopher L.

arXiv.org Artificial Intelligence

In fishery science, harvest management of size-structured stochastic populations is a long-standing and difficult problem. Rectilinear precautionary policies based on biomass and harvesting reference points have now become a standard approach to this problem. While these standard feedback policies are adapted from analytical or dynamic programming solutions assuming relatively simple ecological dynamics, they are often applied to more complicated ecological settings in the real world. In this paper we explore the problem of designing harvest control rules for partially observed, age-structured, spasmodic fish populations using tools from reinforcement learning (RL) and Bayesian optimization. Our focus is on the case of Walleye fisheries in Alberta, Canada, whose highly variable recruitment dynamics have perplexed managers and ecologists. We optimized and evaluated policies using several complementary performance metrics. The main questions we addressed were: 1. How do standard policies based on reference points perform relative to numerically optimized policies? 2. Can an observation of mean fish weight, in addition to stock biomass, aid policy decisions?


Semantic Communication and Control Co-Design for Multi-Objective Correlated Dynamics

Girgis, Abanoub M., Seo, Hyowoon, Bennis, Mehdi

arXiv.org Artificial Intelligence

This letter introduces a machine-learning approach to learning the semantic dynamics of correlated systems with different control rules and dynamics. By leveraging the Koopman operator in an autoencoder (AE) framework, the system's state evolution is linearized in the latent space using a dynamic semantic Koopman (DSK) model, capturing the baseline semantic dynamics. Signal temporal logic (STL) is incorporated through a logical semantic Koopman (LSK) model to encode system-specific control rules. These models form the proposed logical Koopman AE framework that reduces communication costs while improving state prediction accuracy and control performance, showing a 91.65% reduction in communication samples and significant performance gains in simulation.


Scalable Optimal Design of Incremental Volt/VAR Control using Deep Neural Networks

Gupta, Sarthak, Mehrizi-Sani, Ali, Chatzivasileiadis, Spyros, Kekatos, Vassilis

arXiv.org Machine Learning

Volt/VAR control rules facilitate the autonomous operation of distributed energy resources (DER) to regulate voltage in power distribution grids. According to non-incremental control rules, such as the one mandated by the IEEE Standard 1547, the reactive power setpoint of each DER is computed as a piecewise-linear curve of the local voltage. However, the slopes of such curves are upper-bounded to ensure stability. On the other hand, incremental rules add a memory term into the setpoint update, rendering them universally stable. They can thus attain enhanced steady-state voltage profiles. Optimal rule design (ORD) for incremental rules can be formulated as a bilevel program. We put forth a scalable solution by reformulating ORD as training a deep neural network (DNN). This DNN emulates the Volt/VAR dynamics for incremental rules derived as iterations of proximal gradient descent (PGD). Analytical findings and numerical tests corroborate that the proposed ORD solution can be neatly adapted to single/multi-phase feeders.


Deep Learning for Optimal Volt/VAR Control using Distributed Energy Resources

Gupta, Sarthak, Chatzivasileiadis, Spyros, Kekatos, Vassilis

arXiv.org Machine Learning

Given their intermittency, distributed energy resources (DERs) have been commissioned with regulating voltages at fast timescales. Although the IEEE 1547 standard specifies the shape of Volt/VAR control rules, it is not clear how to optimally customize them per DER. Optimal rule design (ORD) is a challenging problem as Volt/VAR rules introduce nonlinear dynamics, require bilinear optimization models, and lurk trade-offs between stability and steady-state performance. To tackle ORD, we develop a deep neural network (DNN) that serves as a digital twin of Volt/VAR dynamics. The DNN takes grid conditions as inputs, uses rule parameters as weights, and computes equilibrium voltages as outputs. Thanks to this genuine design, ORD is reformulated as a deep learning task using grid scenarios as training data and aiming at driving the predicted variables being the equilibrium voltages close to unity. The learning task is solved by modifying efficient deep-learning routines to enforce constraints on rule parameters. In the course of DNN-based ORD, we also review and expand on stability conditions and convergence rates for Volt/VAR rules on single-/multi-phase feeders. To benchmark the optimality and runtime of DNN-based ORD, we also devise a novel mixed-integer nonlinear program formulation. Numerical tests showcase the merits of DNN-based ORD.


Teaching robots to be team players with nature

#artificialintelligence

This en masse behavior by individual organisms can provide separate and collective good, such as improving chances of successful mating propagation or providing security. Now, researchers have harnessed the self-organization skills required to reap the benefits of natural swarms for robotic applications in artificial intelligence, computing, search and rescue, and much more. They published their method on Aug. 3 in Intelligent Computing. "Designing a set of rules that, once executed by a swarm of robots, results in a specific desired behavior is particularly challenging," said corresponding author Marco Dorigo, professor in the artificial intelligence laboratory, named IRIDIA, of the Université Libre de Bruxelles, Belgium. "The behavior of the swarm is not a one-to-one map with simple rules executed by individual robots, but rather results from the complex interactions of many robots executing the same set of rules."


How the Euler Lagrange Equation behaves Part1(Mechanics)

#artificialintelligence

Abstract: Safety-critical control is characterized as ensuring constraint satisfaction for a given dynamical system. Recent developments in zeroing control barrier functions (ZCBFs) have provided a framework for ensuring safety of a superlevel set of a single constraint function. Euler-Lagrange systems represent many real-world systems including robots and vehicles, which must abide by safety-regulations, especially for use in human-occupied environments. These safety regulations include state constraints (position and velocity) and input constraints that must be respected at all times. ZCBFs are valuable for satisfying system constraints for general nonlinear systems, however their construction to satisfy state and input constraints is not straightforward.


Railway Operation Rescheduling System via Dynamic Simulation and Reinforcement Learning

Kubosawa, Shumpei, Onishi, Takashi, Sakahara, Makoto, Tsuruoka, Yoshimasa

arXiv.org Artificial Intelligence

The number of railway service disruptions has been increasing owing to intensification of natural disasters. In addition, abrupt changes in social situations such as the COVID-19 pandemic require railway companies to modify the traffic schedule frequently. Therefore, automatic support for optimal scheduling is anticipated. In this study, an automatic railway scheduling system is presented. The system leverages reinforcement learning and a dynamic simulator that can simulate the railway traffic and passenger flow of a whole line. The proposed system enables rapid generation of the traffic schedule of a whole line because the optimization process is conducted in advance as the training. The system is evaluated using an interruption scenario, and the results demonstrate that the system can generate optimized schedules of the whole line in a few minutes.


A Computational Model of the Institutional Analysis and Development Framework

Montes, Nieves

arXiv.org Artificial Intelligence

The Institutional Analysis and Development (IAD) framework is a conceptual toolbox put forward by Elinor Ostrom and colleagues in an effort to identify and delineate the universal common variables that structure the immense variety of human interactions. The framework identifies rules as one of the core concepts to determine the structure of interactions, and acknowledges their potential to steer a community towards more beneficial and socially desirable outcomes. This work presents the first attempt to turn the IAD framework into a computational model to allow communities of agents to formally perform what-if analysis on a given rule configuration. To do so, we define the Action Situation Language -- or ASL -- whose syntax is hgighly tailored to the components of the IAD framework and that we use to write descriptions of social interactions. ASL is complemented by a game engine that generates its semantics as an extensive-form game. These models, then, can be analyzed with the standard tools of game theory to predict which outcomes are being most incentivized, and evaluated according to their socially relevant properties.


Policy Gradient Reinforcement Learning for Policy Represented by Fuzzy Rules: Application to Simulations of Speed Control of an Automobile

Ishihara, Seiji, Igarashi, Harukazu

arXiv.org Artificial Intelligence

A method of a fusion of fuzzy inference and policy gradient reinforcement learning has been proposed that directly learns, as maximizes the expected value of the reward per episode, parameters in a policy function represented by fuzzy rules with weights. A study has applied this method to a task of speed control of an automobile and has obtained correct policies, some of which control speed of the automobile appropriately but many others generate inappropriate vibration of speed. In general, the policy is not desirable that causes sudden time change or vibration in the output value, and there would be many cases where the policy giving smooth time change in the output value is desirable. In this paper, we propose a fusion method using the objective function, that introduces defuzzification with the center of gravity model weighted stochastically and a constraint term for smoothness of time change, as an improvement measure in order to suppress sudden change of the output value of the fuzzy controller. Then we show the learning rule in the fusion, and also consider the effect by reward functions on the fluctuation of the output value. As experimental results of an application of our method on speed control of an automobile, it was confirmed that the proposed method has the effect of suppressing the undesirable fluctuation in time-series of the output value. Moreover, it was also showed that the difference between reward functions might adversely affect the results of learning.


Kernel-Based Learning for Smart Inverter Control

Garg, Aditie, Jalali, Mana, Kekatos, Vassilis, Gatsis, Nikolaos

arXiv.org Machine Learning

Distribution grids are currently challenged by frequent voltage excursions induced by intermittent solar generation. Smart inverters have been advocated as a fast-responding means to regulate voltage and minimize ohmic losses. Since optimal inverter coordination may be computationally challenging and preset local control rules are subpar, the approach of customized control rules designed in a quasi-static fashion features as a golden middle. Departing from affine control rules, this work puts forth non-linear inverter control policies. Drawing analogies to multi-task learning, reactive control is posed as a kernel-based regression task. Leveraging a linearized grid model and given anticipated data scenarios, inverter rules are jointly designed at the feeder level to minimize a convex combination of voltage deviations and ohmic losses via a linearly-constrained quadratic program. Numerical tests using real-world data on a benchmark feeder demonstrate that nonlinear control rules driven also by a few non-local readings can attain near-optimal performance.